Semantic Similarity Calculation of Chinese Word
نویسندگان
چکیده
منابع مشابه
Semantic Similarity Calculation of Chinese Word
This paper puts forward a two layers computing method to calculate semantic similarity of Chinese word. Firstly, using Latent Dirichlet Allocation (LDA) subject model to generate subject spatial domain. Then mapping word into topic space and forming topic distribution which is used to calculate semantic similarity of word(the first layer computing). Finally, using semantic dictionary"HowNet" to...
متن کاملChinese-English Bilingual Word Semantic Similarity Based on Chinese WordNet
Semantic similarity measurement of multilingual words is a challenging problem in data mining, information extraction, information retrieval, etc. This paper introduces an algorithm to measure the semantic similarity of Chinese-English bilingual words based on Chinese WordNet, an expansion of WordNet in Simplified Chinese. The algorithm not only measures the semantic similarity for Chinese and ...
متن کاملThe Research of Chinese Semantic Similarity Calculation Introduced Punctuations
So far, most Chinese natural language processing neglects the punctuations or oversimplifies their functions. To improve the efficiency of Chinese similarity computing, this paper gives a Chinese similarity computing system model in accordance with the problems of Chinese sentence similarity computation aspect. This model is a combination of punctuations and traditional similarity computing. Co...
متن کاملWord Semantic Similarity Calculation Based on Domain Knowledge and HowNet
Word semantic similarity is the foundation of semantic processing, and is a key issue in many applications. This paper argues that word semantic similarity should associate with domain knowledge, which traditional methods did not take into account. In order to adopt domain knowledge into semantic similarity measurement, this paper proposed a sensitive words sets approach. For this purpose, we a...
متن کاملChinese Word Similarity Measurement
Word similarity computation is a fundamental task for natural language processing. We organize a semantic campaign of Chinese word similarity measurement at NLPCC-ICCPOL 2016. This task provides a benchmark dataset of Chinese word similarity (PKU-500 dataset), including 500 word pairs with their similarity scores. There are 21 teams submitting 24 systems in this campaign. In this paper, we desc...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: International Journal of Advanced Computer Science and Applications
سال: 2014
ISSN: 2156-5570,2158-107X
DOI: 10.14569/ijacsa.2014.050802